S-DFP: shifted dynamic fixed point for quantized deep neural network training
نویسندگان
چکیده
Abstract Recent advances in deep neural networks have achieved higher accuracy with more complex models. Nevertheless, they require much longer training time. To reduce the time, methods using quantized weight, activation, and gradient been proposed. Neural network calculation by integer format improves energy efficiency of hardware for learning Therefore, fixed point However, narrow data representation range degrades accuracy. In this work, we propose a new named shifted dynamic (S-DFP) to prevent degradation training. S-DFP can change adding bias exponent. We evaluated effectiveness on ImageNet task ResNet-34, ResNet-50, ResNet-101 ResNet-152. For example, ResNet-152 is improved from 76.6% conventional 8-bit DFP 77.6% S-DFP.
منابع مشابه
Deep Convolutional Neural Network Inference with Floating-point Weights and Fixed-point Activations
Deep convolutional neural network (CNN) inference requires significant amount of memory and computation, which limits its deployment on embedded devices. To alleviate these problems to some extent, prior research utilize low precision fixed-point numbers to represent the CNN weights and activations. However, the minimum required data precision of fixed-point weights varies across different netw...
متن کاملOvercoming Challenges in Fixed Point Training of Deep Convolutional Networks
It is known that training deep neural networks, in particular, deep convolutional networks, with aggressively reduced numerical precision is challenging. The stochastic gradient descent algorithm becomes unstable in the presence of noisy gradient updates resulting from arithmetic with limited numeric precision. One of the wellaccepted solutions facilitating the training of low precision fixed p...
متن کاملDeep neural network training emphasizing central frames
It is common practice to concatenate several consecutive frames of acoustic features as input of a Deep Neural Network (DNN) for speech recognition. A DNN is trained to map the concatenated frames as a whole to the HMM state corresponding to the center frame while the side frames close to both ends of the concatenated frames and the remaining central frames are treated as equally important. Tho...
متن کامل$S$-metric and fixed point theorem
In this paper, we prove a general fixed point theorem in $textrm{S}$-metric spaces for maps satisfying an implicit relation on complete metric spaces. As applications, we get many analogues of fixed point theorems in metric spaces for $textrm{S}$-metric spaces.
متن کاملMixed Low-precision Deep Learning Inference using Dynamic Fixed Point
We propose a cluster-based quantization method to convert pre-trained full precision weights into ternary weights with minimal impact on the accuracy. In addition we also constrain the activations to 8-bits thus enabling sub 8-bit full integer inference pipeline. Our method uses smaller clusters of N filters with a common scaling factor to minimize the quantization loss, while also maximizing t...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Neural Computing and Applications
سال: 2021
ISSN: ['0941-0643', '1433-3058']
DOI: https://doi.org/10.1007/s00521-021-06821-x